MIP-BOOST: Efficient and Effective L0 Feature Selection for Linear Regression

نویسندگان
چکیده

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Efficient Regularized Regression for Variable Selection with L0 Penalty

Variable (feature, gene, model, which we use interchangeably) selections for regression with high-dimensional BIGDATA have found many applications in bioinformatics, computational biology, image processing, and engineering. One appealing approach is the L0 regularized regression which penalizes the number of nonzero features in the model directly. L0 is known as the most essential sparsity meas...

متن کامل

Efficient Regularized Regression with L0 Penalty for Variable Selection and Network Construction

Variable selections for regression with high-dimensional big data have found many applications in bioinformatics and computational biology. One appealing approach is the L0 regularized regression which penalizes the number of nonzero features in the model directly. However, it is well known that L0 optimization is NP-hard and computationally challenging. In this paper, we propose efficient EM (...

متن کامل

Adaptive Feature Selection: Computationally Efficient Online Sparse Linear Regression under RIP A. Proofs for Realizable Setting

A. Proofs for Realizable Setting Proof of Lemma 3. Let ∆ := ŵ −w∗ be the difference between the true answer and solution to the optimization problem. Let S to be the support of w∗ and let S = [d] \ S be the complements of S. Consider the permutation i1, . . . , id−k of S for which |∆(ij)| ≥ |∆(ij+1)| for all j. That is, the permutation dictated by the magnitude of the entries of ∆ outside of S....

متن کامل

Efficient Learning and Feature Selection in High-Dimensional Regression

We present a novel algorithm for efficient learning and feature selection in high-dimensional regression problems. We arrive at this model through a modification of the standard regression model, enabling us to derive a probabilistic version of the well-known statistical regression technique of backfitting. Using the expectation-maximization algorithm, along with variational approximation metho...

متن کامل

Feature Selection for Regression Problems

Feature subset selection is the process of identifying and removing from a training data set as much irrelevant and redundant features as possible. This reduces the dimensionality of the data and may enable regression algorithms to operate faster and more effectively. In some cases, correlation coefficient can be improved; in others, the result is a more compact, easily interpreted representati...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of Computational and Graphical Statistics

سال: 2021

ISSN: 1061-8600,1537-2715

DOI: 10.1080/10618600.2020.1845184